Constructing unbiased gradient estimators with finite variance for conditional stochastic optimization
نویسندگان
چکیده
We study stochastic gradient descent for solving conditional optimization problems, in which an objective to be minimized is given by a parametric nested expectation with outer taken respect one random variable and inner the other variable. The of such again expressed as expectation, makes it hard standard Monte Carlo estimator unbiased. In this paper, we show under some conditions that multilevel unbiased has finite variance expected computational cost, so theory from (non-nested) directly applies. also discuss special case yet another cost can constructed.
منابع مشابه
Variance Reduction for Stochastic Gradient Optimization
Stochastic gradient optimization is a class of widely used algorithms for training machine learning models. To optimize an objective, it uses the noisy gradient computed from the random data samples instead of the true gradient computed from the entire dataset. However, when the variance of the noisy gradient is large, the algorithm might spend much time bouncing around, leading to slower conve...
متن کاملGuaranteed Sufficient Decrease for Stochastic Variance Reduced Gradient Optimization
In this paper, we propose a novel sufficient decrease technique for stochastic variance reduced gradient descent methods such as SVRG and SAGA. In order to make sufficient decrease for stochastic optimization, we design a new sufficient decrease criterion, which yields sufficient decrease versions of stochastic variance reduction algorithms such as SVRG-SD and SAGA-SD as a byproduct. We introdu...
متن کاملConditional gradient type methods for composite nonlinear and stochastic optimization
In this paper, we present a conditional gradient type (CGT) method for solving a class of composite optimization problems where the objective function consists of a (weakly) smooth term and a strongly convex term. While including this strongly convex term in the subproblems of the classical conditional gradient (CG) method improves its convergence rate for solving strongly convex problems, it d...
متن کاملEnhancing Stochastic Kriging Metamodels with Gradient Estimators
Full terms and conditions of use:
متن کاملEfficient Generalized Conditional Gradient with Gradient Sliding for Composite Optimization
Generalized conditional gradient method has regained increasing research interest as an alternative to another popular proximal gradient method for sparse optimization problems. For particular tasks, its low computation cost of linear subproblem evaluation on each iteration leads to superior practical performance. However, the inferior iteration complexity incurs excess number of gradient evalu...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Mathematics and Computers in Simulation
سال: 2023
ISSN: ['0378-4754', '1872-7166']
DOI: https://doi.org/10.1016/j.matcom.2022.09.012